On Reduction via Determinization of Speech-Recognition Lattices
نویسنده
چکیده
We establish a framework for studying the behavior of automatic speech recognition (ASR) lattices (viewed as automata) undergoing determinization. Using this framework, we provide initial insights into what causes determinization to produce smaller (or bigger) lattices when used in the ASR application. Our results counter the prevailing wisdom that the graph topology underlying an automaton, not the weights on the arcs, governs deterministic expansion. We show that there are graphs that expand solely due to their weights when determinized; i.e., we demonstrate graphs that expand under some weightings yet contract under others. Furthermore, we give evidence that the automata that arise in ASR are either the kind that never expand or else the weight-dependent kind; i.e., we do not nd in ASR any instances of automata that always expand under determinization. Therefore, understanding what causes weight dependence becomes essential to providing tools to avoid deterministic expansion in ASR, and we provide some theoretical results that start to explain this behavior.
منابع مشابه
Speech Recognition with Weighted Finite-state Transducers
This chapter describes a general representation and algorithmic framework for speech recognition based on weighted finite-state transducers. These transducers provide a common and natural representation for major components of speech recognition systems, including hidden Markov models (HMMs), context-dependency models, pronunciation dictionaries, statistical grammars, and word or phone lattices...
متن کاملWeighted finite-state transducers in speech recognition
We survey the weighted finite-state transducer (WFST) approach to speech recognition developed at AT&T over the last several years. We show that WFSTs provide a common and natural representation for HMM models, context-dependency, pronunciation dictionaries, grammars, and alternative recognition outputs. Furthermore, general finite-state operations combine these representations flexibly and eff...
متن کاملEfficient lattice representation and generation
In large-vocabulary, multi-pass speech recognition systems, it is desirable to generate word lattices incorporating a large number of hypotheses while keeping the lattice sizes small. We describe two new techniques for reducing word lattice sizes without eliminating hypotheses. The first technique is an algorithm to reduce the size of non-deterministic bigram word lattices. The algorithm iterat...
متن کاملWeighted Finite - State Transducers inSpeech
We survey the use of weighted nite-state transducers (WFSTs) in speech recognition. We show that WFSTs provide a common and natural representation for HMM models, context-dependency, pronunciation dictionaries , grammars, and alternative recognition outputs. Furthermore, general transducer operations combine these representations exibly and ee-ciently. Weighted determinization and minimization ...
متن کاملWeighted determinization and minimization for large vocabulary speech recognition
ABSTRACT Speech recognition requires solving many space and time problems that can have a critical effect on the overall system performance. We describe the use of two general new algorithms [5] that transform recognition networks into equivalent ones that require much less time and space in large-vocabulary speech recognition. The new algorithms generalize classical automata determinization an...
متن کامل